Improving Generalisation Performance Through Multiobjective Parsimony Enforcement
نویسندگان
چکیده
This paper describes POPE-GP, a system that makes use of the NSGA-II multiobjective evolutionary algorithm as an alternative, parameter-free technique for eliminating program bloat. We test it on a classification problem and find that while vastly reducing program size, the technique does improve generalisation performance. Program Bloat, the phenomenon of ever-increasing program size during a GP run, is a recognised and widespread problem. Traditional techniques to combat program bloat are program size limitations or parsimony pressure (penalty functions). These techniques suffer from a number of problems, in particular their reliance on parameters whose optimal values it is difficult to a priori determine. In this work we study the performance of POPE-GP, a new algorithm that uses the NSGA-II multiobjective algorithm as the basis for parsimony enforcement. We are especially interested in finding out if small solutions generalise better than large solutions. To achieve this, we compare the performance of POPE-GP on a real-world classification problem with that of a GP with more traditional parsimony control and a GP with no control at all, paying particular attention to the performance of solutions on the unseen testing set as a measure of generalisation performance. The Pseudo-Objective Parsimony Enforcement GP (POPE-GP) uses the NSGA-II multiobjective optimisation algorithm [1] as a base for its operation. The two objectives are defined as being the actual objective of the GP run (the fitness) and the size of the program. Once these objectives have been defined, the NSGA-II algorithm attempts to find the Pareto Front for these two objectives. We compared the generalisation performance of classifier programs generated by the POPE-GP algorithm with those generated by a standard GP with a depth limit of eight and one with no limits at all. We used the Wisconsin Breast Cancer Database1, which has been widely used as a testbed for classification. We divided the 699 instances in the data set randomly into training and testing sets, so that 70% (479 instances) of the data made up the training set and the remaining 30% (204 instances) constituted the testing set. We used the RMITGP2 GP programming library with strongly-typed GP. The fitness of an individual was taken to be the gross classification error – ie. the number of instances in the training set that are misclassified. 1 http://www.ics.uci.edu/∼mlearn/MLSummary.html 2 http://yallara.cs.rmit.edu.au/∼dylanm/rmitgp.html K. Deb et al. (Eds.): GECCO 2004, LNCS 3103, pp. 702–703, 2004. c © Springer-Verlag Berlin Heidelberg 2004 Improving Generalisation Performance 703 Table 1. (a) End-of-run average values for the algorithms tested. (b) Mean classification accuracy on the testing set. Algorithm AvDepth AvFitness AvSize BestDepth BestFitness BestSize POPE-GP (500) 6.71 0.9586 18.77 9.40 0.9865 31.50 POPE-GP (50) 5.40 0.9467 13.12 8.00 0.9811 23.20 Depth-Limited (500) 7.99 0.9388 300.79 8.00 0.9845 282.72 Depth-Limited (50) 7.97 0.9175 270.27 7.86 0.9753 261.06 No Parsimony Pressure (500) 33.99 0.9658 1266.01 21.30 0.9858 691.56
منابع مشابه
Comparing Different Operators and Models to Improve a Multiobjective Artificial Bee Colony Algorithm for Inferring Phylogenies
Maximum parsimony and maximum likelihood approaches to phylogenetic reconstruction were proposed with the aim of describing the evolutionary history of species by using different optimality principles. These discrepant points of view can lead to situations where discordant topologies are inferred from a same dataset. In recent years, research efforts in Phylogenetics try to apply multiobjective...
متن کاملAccuracy, Parsimony, and Generality in Evolutionary Learning Systems via Multiobjective Selection
Learning systems (also known as Pittsburgh learning classifier systems) need to balance accuracy and parsimony for evolving high quality general hypotheses. The evolutionary learning process used in learning systems is based on using a set of training instances that sample the target concept to be learned. Thus, the the learning process may overfit the learned hypothesis to the given set of tra...
متن کاملMultiobjective Optimization of Classifiers by Means of 3-D Convex Hull Based Evolutionary Algorithm
Finding a good classifier is a multiobjective optimization problem with different error rates and the costs to be minimized. The receiver operating characteristic is widely used in the machine learning community to analyze the performance of parametric classifiers or sets of Pareto optimal classifiers. In order to directly compare two sets of classifiers the area (or volume) under the convex hu...
متن کاملOn Generalisation of Machine Learning with Neural-Evolutionary Computations
Generalisation is a non-trivial problem in machine learning and more so with neural networks which have the capabilities of inducing varying degrees of freedom. It is influenced by many factors in network design, such as network size, initial conditions, learning rate, weight decay factor, pruning algorithms, and many more. In spite of continuous research efforts we could not arrive at a practi...
متن کاملXergy analysis and multiobjective optimization of a biomass gasification-based multigeneration system
Biomass gasification is the process of converting biomass into a combustible gas suitable for use in boilers, engines, and turbines to produce combined cooling, heat, and power. This paper presents a detailed model of a biomass gasification system and designs a multigeneration energy system that uses the biomass gasification process for generating combined cooling, heat, and electricity. Energy...
متن کامل